5 research outputs found
Affective Behaviour Analysis of On-line User Interactions: Are On-line Support Groups more Therapeutic than Twitter?
The increase in the prevalence of mental health problems has coincided with a
growing popularity of health related social networking sites. Regardless of
their therapeutic potential, On-line Support Groups (OSGs) can also have
negative effects on patients. In this work we propose a novel methodology to
automatically verify the presence of therapeutic factors in social networking
websites by using Natural Language Processing (NLP) techniques. The methodology
is evaluated on On-line asynchronous multi-party conversations collected from
an OSG and Twitter. The results of the analysis indicate that therapeutic
factors occur more frequently in OSG conversations than in Twitter
conversations. Moreover, the analysis of OSG conversations reveals that the
users of that platform are supportive, and interactions are likely to lead to
the improvement of their emotional state. We believe that our method provides a
stepping stone towards automatic analysis of emotional states of users of
online platforms. Possible applications of the method include provision of
guidelines that highlight potential implications of using such platforms on
users' mental health, and/or support in the analysis of their impact on
specific individuals
Automatically Predicting User Ratings for Conversational Systems
Automatic evaluation models for open-domain conversational agents either correlate poorly with human judgment or require expensive annotations on top of conversation scores. In this work we investigate the feasibility of learning evaluation models without relying on any further annotations besides conversation-level human ratings. We use a dataset of rated (1-5) open domain spoken conversations between the conversational agent Roving Mind (competing in the Amazon Alexa Prize Challenge 2017) and Amazon Alexa users. First, we assess the complexity of the task by asking two experts to re-annotate a sample of the dataset and observe that the subjectivity of user ratings yields a low upper-bound. Second, through an analysis of the entire dataset we show that automatically extracted features such as user sentiment, Dialogue Acts and conversation length have significant, but low correlation with user ratings. Finally, we report the results of our experiments exploring different combinations of these features to train automatic dialogue evaluation models. Our work suggests that predicting subjective user ratings in open domain conversations is a challenging task.I modelli stato dellâarte per la valutazione automatica di agenti conversazionali open-domain hanno una scarsa correlazione con il giudizio umano oppure richiedono costose annotazioni oltre al punteggio dato alla conversazione. In questo lavoro investighiamo la possibilitĂ di apprendere modelli di valutazione attraverso il solo utilizzo di punteggi umani dati allâintera conversazione. Il corpus utilizzato è composto da conversazioni parlate open-domain tra lâagente conversazionale Roving Mind (parte della competizione Amazon Alexa Prize 2017) e utenti di Amazon Alexa valutate con punteggi da 1 a 5. In primo luogo, valutiamo la complessitĂ del task assegnando a due esperti il compito di riannotare una parte del corpus e osserviamo come esso risulti complesso perfino per annotatori umani data la sua soggettivitĂ . In secondo luogo, tramite unâanalisi condotta sullâintero corpus mostriamo come features estratte automaticamente (sentimento dellâutente, Dialogue Acts e lunghezza della conversazione) hanno bassa, ma significativa correlazione con il giudizio degli utenti. Infine, riportiamo i risultati di esperimenti volti a esplorare diverse combinazioni di queste features per addestrare modelli di valutazione automatica del dialogo. Questo lavoro mostra la difficoltĂ del predire i giudizi soggettivi degli utenti in conversazioni senza un task specifico
Proceedings of the Fifth Italian Conference on Computational Linguistics CLiC-it 2018
On behalf of the Program Committee, a very warm welcome to the Fifth Italian Conference on Computational Linguistics (CLiC-Ââit 2018). This edition of the conference is held in Torino. The conference is locally organised by the University of Torino and hosted into its prestigious main lecture hall âCavallerizza Realeâ. The CLiC-Ââit conference series is an initiative of the Italian Association for Computational Linguistics (AILC) which, after five years of activity, has clearly established itself as the premier national forum for research and development in the fields of Computational Linguistics and Natural Language Processing, where leading researchers and practitioners from academia and industry meet to share their research results, experiences, and challenges